Goto

Collaborating Authors

 threshold function


Smoothed Agnostic Learning of Halfspaces over the Hypercube

Kou, Yiwen, Meka, Raghu

arXiv.org Machine Learning

Agnostic learning of Boolean halfspaces is a fundamental problem in computational learning theory, but it is known to be computationally hard even for weak learning. Recent work [CKKMK24] proposed smoothed analysis as a way to bypass such hardness, but existing frameworks rely on additive Gaussian perturbations, making them unsuitable for discrete domains. We introduce a new smoothed agnostic learning framework for Boolean inputs, where perturbations are modeled via random bit flips. This defines a natural discrete analogue of smoothed optimality generalizing the Gaussian case. Under strictly subexponential assumptions on the input distribution, we give an efficient algorithm for learning halfspaces in this model, with runtime and sample complexity approximately n raised to a poly(1/(sigma * epsilon)) factor. Previously, such algorithms were known only with strong structural assumptions for the discrete hypercube, for example, independent coordinates or symmetric distributions. Our result provides the first computationally efficient guarantee for smoothed agnostic learning of halfspaces over the Boolean hypercube, bridging the gap between worst-case intractability and practical learnability in discrete settings.


Influence Maximization with \varepsilon -Almost Submodular Threshold Functions

Neural Information Processing Systems

Influence maximization is the problem of selecting $k$ nodes in a social network to maximize their influence spread. The problem has been extensively studied but most works focus on the submodular influence diffusion models. In this paper, motivated by empirical evidences, we explore influence maximization in the non-submodular regime. In particular, we study the general threshold model in which a fraction of nodes have non-submodular threshold functions, but their threshold functions are closely upper-and lower-bounded by some submodular functions (we call them $\varepsilon$-almost submodular). We first show a strong hardness result: there is no $1/n^{\gamma/c}$ approximation for influence maximization (unless P = NP) for all networks with up to $n^{\gamma}$ $\varepsilon$-almost submodular nodes, where $\gamma$ is in (0,1) and $c$ is a parameter depending on $\varepsilon$. This indicates that influence maximization is still hard to approximate even though threshold functions are close to submodular.



On Neuronal Capacity

Pierre Baldi, Roman Vershynin

Neural Information Processing Systems

We define the capacity of a learning machine to be the logarithm of the number (or volume) of the functions it can implement.



A Section 3 details

Neural Information Processing Systems

We prove Lemma 4. Lemma 4 (restated). In this section, the proofs omitted in Section 4 are presented. We first define sub-trees . Next we prove a helper lemma. 's whose summation is less than Now we are ready to prove Theorem 8. Theorem 8 We begin with the multi-class setting.




On Neuronal Capacity

Pierre Baldi, Roman Vershynin

Neural Information Processing Systems

We define the capacity of a learning machine to be the logarithm of the number (or volume) of the functions it can implement.


Asymptotic analysis of cooperative censoring policies in sensor networks

Fernandez-Bes, Jesus, Arroyo-Valles, Rocío, Cid-Sueiro, Jesús

arXiv.org Artificial Intelligence

The problem of cooperative data censoring in battery-powered multihop sensor networks is analyzed in this paper. We are interested in scenarios where nodes generate messages (which are related to the sensor measurements) that can be graded with some importance value. Less important messages can be censored in order to save energy for later communications. The problem is modeled using a joint Markov Decision Process of the whole network dynamics, and a theoretically optimal censoring policy, which maximizes a long-term reward, is found. Though the optimal censoring rules are computationally prohibitive, our analysis suggests that, under some conditions, they can be approximated by a finite collection of constant-threshold rules. A centralized algorithm for the computation of these thresholds is proposed. The experimental simulations show that cooperative censoring policies are energy-efficient, and outperform other non-cooperative schemes.